首页> 外文OA文献 >Approximate Clustering without the Approximation
【2h】

Approximate Clustering without the Approximation

机译:没有近似的近似聚类

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Approximation algorithms for clustering points in metric spaces is a flourishing area of research, with much research effort spent on getting a better understanding of the approximation guarantees possible for many objective functions such as k-median, k-means, and min-sum clustering.This quest for better approximation algorithms is further fueled by the implicit hope that these better approximations also give us more accurate clusterings. E.g., for many problems such as clustering proteins by function, or clustering images by subject, there is some unknown “correct” target clustering and the implicit hope is that approximately optimizing these objective functions will in fact produce a clustering that is close (in symmetric difference) to the truth.In this paper, we show that if we make this implicit assumption explicit—that is, if we assume that any c-approximation to the given clustering objective F is ∈-close to the target—then we can produce clusterings that are O(∈)-close to the target, even for values c for which obtaining a c-approximation is NP-hard. In particular, for k-median and k-means objectives, we show that we can achieve this guarantee for any constant c \u3e 1, and for min-sum objective we can do this for any constant c \u3e 2.Our results also highlight a somewhat surprising conceptual difference between assuming that the optimal solution to, say, the k-median objective is ∈-close to the target, and assuming that any approximately optimal solution is ǫ-close to the target, even for approximation factor say c = 1.01. In the former case, the problem of finding a solution that is O(∈)-close to the target remains computationally hard, and yet for the latter we have an efficient algorithm.
机译:用于度量空间中的聚类点的近似算法是一个蓬勃发展的研究领域,大量的研究工作花费在更好地理解许多目标函数(如k中值,k均值和最小和聚类)可能的近似保证上。对更好的近似算法的追求进一步被这些更好的近似也给我们更准确的聚类的隐含希望所激发。例如,对于许多问题,例如按功能对蛋白质进行聚类或按主题对图像进行聚类,存在一些未知的“正确”目标聚类,并且隐含的希望是,对这些目标函数进行近似优化实际上会产生紧密的聚类(对称在本文中,我们证明了,如果我们使该隐式假设变得明确,即假设给定聚类目标F的任何c逼近都接近目标,则可以产生O(∈)接近目标的聚类,即使对于获得c近似值是NP困难的c值也是如此。特别是,对于k中位数和k均值目标,我们表明我们可以对任何常数c \ u3e 1达到此保证,对于最小和目标,我们可以对任何常数c \ u3e 2做到这一点。我们的结果强调一个在概念上有些令人惊讶的区别,即假设假设k个中值目标的最优解接近目标,而假设近似最优解近似于目标,即使近似值c = 1.01。在前一种情况下,找到一个与目标接近O(ε)的解的问题在计算上仍然很困难,但对于后者,我们有一个有效的算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号